7,015 research outputs found
"Mental Rotation" by Optimizing Transforming Distance
The human visual system is able to recognize objects despite transformations
that can drastically alter their appearance. To this end, much effort has been
devoted to the invariance properties of recognition systems. Invariance can be
engineered (e.g. convolutional nets), or learned from data explicitly (e.g.
temporal coherence) or implicitly (e.g. by data augmentation). One idea that
has not, to date, been explored is the integration of latent variables which
permit a search over a learned space of transformations. Motivated by evidence
that people mentally simulate transformations in space while comparing
examples, so-called "mental rotation", we propose a transforming distance.
Here, a trained relational model actively transforms pairs of examples so that
they are maximally similar in some feature space yet respect the learned
transformational constraints. We apply our method to nearest-neighbour problems
on the Toronto Face Database and NORB
Explaining the Unexplained: A CLass-Enhanced Attentive Response (CLEAR) Approach to Understanding Deep Neural Networks
In this work, we propose CLass-Enhanced Attentive Response (CLEAR): an
approach to visualize and understand the decisions made by deep neural networks
(DNNs) given a specific input. CLEAR facilitates the visualization of attentive
regions and levels of interest of DNNs during the decision-making process. It
also enables the visualization of the most dominant classes associated with
these attentive regions of interest. As such, CLEAR can mitigate some of the
shortcomings of heatmap-based methods associated with decision ambiguity, and
allows for better insights into the decision-making process of DNNs.
Quantitative and qualitative experiments across three different datasets
demonstrate the efficacy of CLEAR for gaining a better understanding of the
inner workings of DNNs during the decision-making process.Comment: Accepted at Computer Vision and Patter Recognition Workshop (CVPR-W)
on Explainable Computer Vision, 201
Opening the Black Box of Financial AI with CLEAR-Trade: A CLass-Enhanced Attentive Response Approach for Explaining and Visualizing Deep Learning-Driven Stock Market Prediction
Deep learning has been shown to outperform traditional machine learning
algorithms across a wide range of problem domains. However, current deep
learning algorithms have been criticized as uninterpretable "black-boxes" which
cannot explain their decision making processes. This is a major shortcoming
that prevents the widespread application of deep learning to domains with
regulatory processes such as finance. As such, industries such as finance have
to rely on traditional models like decision trees that are much more
interpretable but less effective than deep learning for complex problems. In
this paper, we propose CLEAR-Trade, a novel financial AI visualization
framework for deep learning-driven stock market prediction that mitigates the
interpretability issue of deep learning methods. In particular, CLEAR-Trade
provides a effective way to visualize and explain decisions made by deep stock
market prediction models. We show the efficacy of CLEAR-Trade in enhancing the
interpretability of stock market prediction by conducting experiments based on
S&P 500 stock index prediction. The results demonstrate that CLEAR-Trade can
provide significant insight into the decision-making process of deep
learning-driven financial models, particularly for regulatory processes, thus
improving their potential uptake in the financial industry
- …